# Disentangled Attention
Erlangshen DeBERTa V2 320M Chinese
Apache-2.0
Chinese pre-trained language model based on DeBERTa-v2 architecture with 320 million parameters, excelling in natural language understanding tasks
Large Language Model
Transformers Chinese

E
IDEA-CCNL
186
13
Deberta V3 Base
MIT
DeBERTaV3 is an improved pre-trained language model based on DeBERTa, which enhances efficiency through gradient-disentangled embedding sharing in ELECTRA-style pretraining and excels in natural language understanding tasks.
Large Language Model English
D
microsoft
1.6M
316
Featured Recommended AI Models